Description
The Artificial General Intelligence team is responsibly advancing the company’s generative AI technologies, including Amazon’s most expansive multimodal Large Language Models. Our inference engines power these initiatives.
You will leverage advanced hardware, innovative software architecture, and distributed computing techniques to enable breakthrough research and product development across the company.
Your innovations in the inference runtime will be instrumental in establishing Amazon as the market leader in enterprise AI solutions.
Key job responsibilities
You will develop, improve and release our cutting-edge inference engines.
You will lead our efforts to build the best inference performance on custom AWS Trainium and Inferentia silicon and the Trn1, Inf1/2 servers. Strong software development (Python and C++) and Machine Learning knowledge (Text and Multimodal) are both critical to this role.
You will understand current and future directions of ML framework development, with a focus on enabling the fastest and most price-performant inference.
About the team
Our mission is to build the fastest and most price-performant inference engines serving AGI's best-in-class models for Amazon's customers.
We are a Science and Engineering team working on the cutting edge of inference.
We are interested in tackling the hardest and most impactful problems in AI inference. We explore inference-aware architectures, and compiler, kernel and runtime improvements to serve AI models of increasing size and performance.
Basic Qualifications
5+ years of non-internship professional software development experience
5+ years of programming with at least one software programming language experience
5+ years of leading design or architecture (design patterns, reliability and scaling) of new and existing systems experience
5+ years of full software development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations experience
Experience as a mentor, tech lead or leading an engineering team
Fundamentals of machine learning, prevalent model architectures including Transformers, familiarity of training and inference lifecycles, work experience with inference frameworks and inference optimizations.
Preferred Qualifications
- Bachelor's degree in computer science or equivalent
Amazon is committed to a diverse and inclusive workplace. Amazon is an equal opportunity employer and does not discriminate on the basis of race, national origin, gender, gender identity, sexual orientation, protected veteran status, disability, age, or other legally protected status. For individuals with disabilities who would like to request an accommodation, please visit https://www.amazon.jobs/en/disability/us.
Our compensation reflects the cost of labor across several US geographic markets. The base pay for this position ranges from $151,300/year in our lowest geographic market up to $261,500/year in our highest geographic market. Pay is based on a number of factors including market location and may vary depending on job-related knowledge, skills, and experience. Amazon is a total compensation company. Dependent on the position offered, equity, sign-on payments, and other forms of compensation may be provided as part of a total compensation package, in addition to a full range of medical, financial, and/or other benefits. For more information, please visit https://www.aboutamazon.com/workplace/employee-benefits. This position will remain posted until filled. Applicants should apply via our internal or external career site.